Personnel
Overall Objectives
Research Program
Application Domains
Highlights of the Year
New Software and Platforms
New Results
Bilateral Contracts and Grants with Industry
Partnerships and Cooperations
Dissemination
Bibliography
XML PDF e-pub
PDF e-Pub


Section: New Results

Tracklets Pre-Processing for Signature Computation in the Context of Multi-Shot Person Re-Identification

Participants : Salwa Baabou, Furqan Muhammad Khan, Thi Lan Anh Nguyen, Thanh Hung Nguyen, François Brémond.

keywords: Person Re-Identification (Re-ID), tracklet, signature representation.

Tracklets pre-processing/representation

The person Re-Identification (Re-ID) system is divided into two steps: i) constructing a person's appearance signature by extracting feature representations which should be robust against pose variations, illumination changes and occlusions and ii) Establishing the correspondence/matching between feature representations of probe and gallery by learning similarity metrics or ranking functions (see Figure 10). However, appearance based person Re-ID is a challenging task due to disparities of human bodies and visual ambiguities across different cameras. Therefore, we focus on how to pre-process tracklets to compute the signature and represent it for Multi-shot person Re-ID to handle high appearance variance and occlusions.

Figure 10. Re-ID Diagram
IMG/salwa-fig-1.png

We have worked on CHU Nice dataset. First, we used the SSD detector  [91] to get the detection results. Second, inspired by the Multi-Object Tracking in [35], we extracted the tracklets of persons and then we used these tracklets to compute the signatures of individuals based on Part Appearance Mixture PAM approach [31].

Figure  11 shows a visualizaton of a sample from CHU Nice dataset of the detection results.

Figure 11. A visualization of a sample from CHU Nice dataset of the Detection Results
IMG/salwa-Image.png IMG/Imagedepth.png
(a) (b)

Figure  12 presents a visualization of a sample of consecutive frames from CHU Nice dataset of the tracking results.

Figure 12. A visualization of a sample of frames from CHU Nice dataset of the Tracking Results
IMG/salwa-Image15.png IMG/salwa-Image16.png IMG/salwa-Image17.png IMG/salwa-Image18.png IMG/salwa-Image19.png IMG/salwa-Image20.png
(a) (b) (c) (d) (e) (f)

Figure  13 shows a representation of some samples of tracklets of a person from CHU Nice dataset.

Figure 13. A sample of tracklets representation from CHU Nice Dataset
IMG/salwa-Image3.png IMG/salwa-Image4.png IMG/salwa-Image5.png IMG/salwa-Image6.png IMG/salwa-Image7.png IMG/salwa-Image02.png IMG/salwa-Image03.png IMG/salwa-Image04.png IMG/salwa-Image05.png IMG/salwa-Image06.png
(a) (b) (c) (d) (e) (f) (g) (h) (i) (j)

Experimental results

CHU Nice dataset

It is an RGB-D dataset (RGB+Depth) and Collected in the hospital of Nice (CHU) in Nice, France. Most of the people recruited for this dataset were elderly people, aged 65 and above, of both genders. It contains 615 videos with 149365 frames acquired from 2 cameras: one in the corridor and another camera in the room.

Table 4 shows the recognition rate (%) at different ranks (rank-1, 5, 10, 20) of a baseline method LOMO+XQDA  [89], PAM-LOMO+XQDA and PAM-LOMO+KISSME on CHU Nice dataset. From the above experiments, we notice that PAM-LOMO+KISSME achieves good performance on three datasets; 81.8% rank-1 recognition rate. This shows that our adaptation of feature descriptor LOMO  [89] and metric learning KISSME  [84] to PAM representation is effective.

Limitations

As shown in Figure  14, a visualization of a selected samples from CHU Nice dataset of PAM signature representation is presented. Indeed, we visualize each GMM component by constructing a composite image. Given appearance descriptor, we compute the likelihood of an image belonging to a model component and then by summing images of corresponding person weighted by their likelihood we generate the composite image. We can say that our signature representation is able to cater variance in person's pose and orientation as well as illumination, it deals also with occlusions and is able to reduce effect of background. However, we can notice that this PAM signature representation presents some limitations, specially on our own dataset CHU Nice, which can affect the quality of our signature representation (see Figure 5). Among these challenging problems, we can cite:

We are actually working to improve the PAM signature representation by using the skeleton and extracting the pose machines from our dataset CHU Nice.

Figure 14. A visualization of selected samples of signature representation from CHU Nice Dataset: the first image is the input image used to learn appearance model. It's followed by the composite images, one for each component of the GMM mixture. Optimal number of GMM components for each appearance model varies between persons. GMM components focus on different pose and orientation of the person.
IMG/HOG33w.png IMG/HOG33w00.png IMG/HOG33w001.png IMG/HOG33w01.png IMG/HOG33w011.png IMG/HOG33w02.png IMG/HOG33w021.png
(a) (b) (c) (d) (e) (f) (g)